Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Review of pre-trained models for natural language processing tasks
LIU Ruiheng, YE Xia, YUE Zengying
Journal of Computer Applications    2021, 41 (5): 1236-1246.   DOI: 10.11772/j.issn.1001-9081.2020081152
Abstract893)      PDF (1296KB)(3013)       Save
In recent years, deep learning technology has developed rapidly. In Natural Language Processing (NLP) tasks, with text representation technology rising from the word level to the document level, the unsupervised pre-training method using a large-scale corpus has been proved to be able to effectively improve the performance of models in downstream tasks. Firstly, according to the development of text feature extraction technology, typical models were analyzed from word level and document level. Secondly, the research status of the current pre-trained models was analyzed from the two stages of pre-training target task and downstream application, and the characteristics of the representative models were summed up. Finally, the main challenges faced by the development of pre-trained models were summarized and the prospects were proposed.
Reference | Related Articles | Metrics